Parameter-Efficient Deep Neural Networks With Bilinear Projections
نویسندگان
چکیده
Recent research on deep neural networks (DNNs) has primarily focused improving the model accuracy. Given a proper learning framework, it is generally possible to increase depth or layer width achieve higher level of However, huge number parameters imposes more computational and memory usage overhead leads parameter redundancy. In this paper, we address redundancy problem in DNNs by replacing conventional full projections with bilinear projections. For fully-connected $D$ input nodes output nodes, applying projection can reduce space complexity from $\mathcal{O}(D^2)$ $\mathcal{O}(2D)$, achieving sub-linear size. structured lower freedom degree compared projection, causing under-fitting problem. So simply scale up mapping size increasing channels, which keep even boosts This makes very parameter-efficient handy deploy such models mobile systems limitations. Experiments four benchmark datasets show that proposed accuracies than DNNs, while significantly reduces
منابع مشابه
ProjectionNet: Learning Efficient On-Device Deep Networks Using Neural Projections
Deep neural networks have become ubiquitous for applications related to visual recognition and language understanding tasks. However, it is often prohibitive to use typical neural network models on devices like mobile phones or smart watches since the model sizes are huge and cannot fit in the limited memory available on such devices. While these devices could make use of machine learning model...
متن کاملSRINIVAS, BABU: DATA-FREE PARAMETER PRUNING FOR DEEP NEURAL NETWORKS 1 Data-free Parameter Pruning for Deep Neural Networks
Deep Neural nets (NNs) with millions of parameters are at the heart of many stateof-the-art computer vision systems today. However, recent works have shown that much smaller models can achieve similar levels of performance. In this work, we address the problem of pruning parameters in a trained NN model. Instead of removing individual weights one at a time as done in previous works, we remove o...
متن کاملEfficient Inferencing of Compressed Deep Neural Networks
Large number of weights in deep neural networks makes the models difficult to be deployed in low memory environments such as, mobile phones, IOT edge devices as well as “inferencing as a service” environments on cloud. Prior work has considered reduction in the size of the models, through compression techniques like pruning, quantization, Huffman encoding etc. However, efficient inferencing usi...
متن کاملEfficient Model Averaging for Deep Neural Networks
Large neural networks trained on small datasets are increasingly prone to overfitting. Traditional machine learning methods can reduce overfitting by employing bagging or boosting to train several diverse models. For large neural networks, however, this is prohibitively expensive. To address this issue, we propose a method to leverage the benefits of ensembles without explicitely training sever...
متن کاملData-free Parameter Pruning for Deep Neural Networks
Deep Neural nets (NNs) with millions of parameters are at the heart of many stateof-the-art computer vision systems today. However, recent works have shown that much smaller models can achieve similar levels of performance. In this work, we address the problem of pruning parameters in a trained NN model. Instead of removing individual weights one at a time as done in previous works, we remove o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE transactions on neural networks and learning systems
سال: 2021
ISSN: ['2162-237X', '2162-2388']
DOI: https://doi.org/10.1109/tnnls.2020.3016688